The process of implementing a damage detection and characterization strategy for engineering structures is referred to as Structural Health Monitoring (SHM). Here damage is defined as changes to the material and/or geometric properties of a structural system, including changes to the boundary conditions and system connectivity, which adversely affect the system’s performance. The SHM process involves the observation of a system over time using periodically sampled dynamic response measurements from an array of sensors, the extraction of damage-sensitive features from these measurements, and the statistical analysis of these features to determine the current state of system health. For long term SHM, the output of this process is periodically updated information regarding the ability of the structure to perform its intended function in light of the inevitable aging and degradation resulting from operational environments. After extreme events, such as earthquakes or blast loading, SHM is used for rapid condition screening and aims to provide, in near real time, reliable information regarding the integrity of the structure.[1]
Contents |
Qualitative and non-continuous methods have long been used to evaluate structures for their capacity to serve their intended purpose. Since the beginning of the 19th century, railroad wheel-tappers have used the sound of a hammer striking the train wheel to evaluate if damage was present.[2] In rotating machinery, vibration monitoring has been used for decades as a performance evaluation technique.[1] In the last ten to fifteen years, SHM technologies have emerged creating an exciting new field within various branches of engineering. Academic conferences and scientific journals have been established during this time that specifically focus on SHM.[2] These technologies are currently becoming increasingly common.
The SHM problem can be addressed in the context of a statistical pattern recognition paradigm.[3][4] This paradigm can be broken down into four parts: (1) Operational Evaluation, (2) Data Acquisition and Cleansing, (3) Feature Extraction and Data Compression, and (4) Statistical Model Development for Feature Discrimination. When one attempts to apply this paradigm to data from real world structures, it quickly becomes apparent that the ability to cleanse, compress, normalize and fuse data to account for operational and environmental variability is a key implementation issue when addressing Parts 2-4 of this paradigm. These processes can be implemented through hardware or software and, in general, some combination of these two approaches will be used.
Operational evaluation attempts to answer four questions regarding the implementation of a damage identification capability:
Operational evaluation begins to set the limitations on what will be monitored and how the monitoring will be accomplished. This evaluation starts to tailor the damage identification process to features that are unique to the system being monitored and tries to take advantage of unique features of the damage that is to be detected.
The data acquisition portion of the SHM process involves selecting the excitation methods, the sensor types, number and locations, and the data acquisition/storage/transmittal hardware. Again, this process will be application specific. Economic considerations will play a major role in making these decisions. The intervals at which data should be collected is another consideration that must be addressed.
Because data can be measured under varying conditions, the ability to normalize the data becomes very important to the damage identification process. As it applies to SHM, data normalization is the process of separating changes in sensor reading caused by damage from those caused by varying operational and environmental conditions. One of the most common procedures is to normalize the measured responses by the measured inputs. When environmental or operational variability is an issue, the need can arise to normalize the data in some temporal fashion to facilitate the comparison of data measured at similar times of an environmental or operational cycle. Sources of variability in the data acquisition process and with the system being monitored need to be identified and minimized to the extent possible. In general, not all sources of variability can be eliminated. Therefore, it is necessary to make the appropriate measurements such that these sources can be statistically quantified. Variability can arise from changing environmental and test conditions, changes in the data reduction process, and unit-to-unit inconsistencies.
Data cleansing is the process of selectively choosing data to pass on to or reject from the feature selection process. The data cleansing process is usually based on knowledge gained by individuals directly involved with the data acquisition. As an example, an inspection of the test setup may reveal that a sensor was loosely mounted and, hence, based on the judgment of the individuals performing the measurement, this set of data or the data from that particular sensor may be selectively deleted from the feature selection process. Signal processing techniques such as filtering and re-sampling can also be thought of as data cleansing procedures.
Finally, the data acquisition, normalization, and cleansing portion of SHM process should not be static. Insight gained from the feature selection process and the statistical model development process will provide information regarding changes that can improve the data acquisition process.
The area of the SHM process that receives the most attention in the technical literature is the identification of data features that allows one to distinguish between the undamaged and damaged structure. Inherent in this feature selection process is the condensation of the data. The best features for damage identification are, again, application specific.
One of the most common feature extraction methods is based on correlating measured system response quantities, such a vibration amplitude or frequency, with the first-hand observations of the degrading system. Another method of developing features for damage identification is to apply engineered flaws, similar to ones expected in actual operating conditions, to systems and develop an initial understanding of the parameters that are sensitive to the expected damage. The flawed system can also be used to validate that the diagnostic measurements are sensitive enough to distinguish between features identified from the undamaged and damaged system. The use of analytical tools such as experimentally-validated finite element models can be a great asset in this process. In many cases the analytical tools are used to perform numerical experiments where the flaws are introduced through computer simulation. Damage accumulation testing, during which significant structural components of the system under study are degraded by subjecting them to realistic loading conditions, can also be used to identify appropriate features. This process may involve induced-damage testing, fatigue testing, corrosion growth, or temperature cycling to accumulate certain types of damage in an accelerated fashion. Insight into the appropriate features can be gained from several types of analytical and experimental studies as described above and is usually the result of information obtained from some combination of these studies.
The operational implementation and diagnostic measurement technologies needed to perform SHM produce more data than traditional uses of structural dynamics information. A condensation of the data is advantageous and necessary when comparisons of many feature sets obtained over the lifetime of the structure are envisioned. Also, because data will be acquired from a structure over an extended period of time and in an operational environment, robust data reduction techniques must be developed to retain feature sensitivity to the structural changes of interest in the presence of environmental and operational variability. To further aid in the extraction and recording of quality data needed to perform SHM, the statistical significance of the features should be characterized and used in the condensation process.
The portion of the SHM process that has received the least attention in the technical literature is the development of statistical models for discrimination between features from the undamaged and damaged structures. Statistical model development is concerned with the implementation of the algorithms that operate on the extracted features to quantify the damage state of the structure. The algorithms used in statistical model development usually fall into three categories. When data are available from both the undamaged and damaged structure, the statistical pattern recognition algorithms fall into the general classification referred to as supervised learning. Group classification and regression analysis are categories of supervised learning algorithms. Unsupervised learning refers to algorithms that are applied to data not containing examples from the damaged structure. Outlier or novelty detection is the primary class of algorithms applied in unsupervised learning applications. All of the algorithms analyze statistical distributions of the measured or derived features to enhance the damage identification process.
Based on the extensive literature that has developed on SHM over the last 20 years, it can be argued that this field has matured to the point where several fundamental axioms, or general principles, have emerged.[5] The axioms are listed as follows:
SHM System's elements include:
An example of this technology is embedding sensors in structures like bridges and aircraft. These sensors provide real time monitoring of various structural changes like stress and strain. In the case of civil engineering structures, the data provided by the sensors is usually transmitted to a remote data acquisition centres. With the aid of modern technology, real time control of structures (Active Structural Control) based on the information of sensors is possible
The Wind and Structural Health Monitoring System (WASHMS) is a sophisticated bridge monitoring system, costing US$1.3 million, used by the Hong Kong Highways Department to ensure road user comfort and safety of the Tsing Ma, Ting Kau, and Kap Shui Mun bridges that run between Hong Kong and the Hong Kong Airport.
In order to oversee the integrity, durability and reliability of the bridges, WASHMS has four different levels of operation: sensory systems, data acquisition systems, local centralised computer systems and global central computer system.
The sensory system consists of approximately 900 sensors and their relevant interfacing units. With more than 350 sensors on the Tsing Ma bridge, 350 on Ting Kau and 200 on Kap Shui Mun, the structural behaviour of the bridges is measured 24 hours a day, seven days a week.
The sensors include accelerometers, strain gauges, displacement transducers, level sensing stations, anemometers, temperature sensors and dynamic weight-in-motion sensors. They measure everything from tarmac temperature and strains in structural members to wind speed and the deflection and rotation of the kilometres of cables and any movement of the bridge decks and towers.
These sensors are the early warning system for the bridges, providing the essential information that help the Highways Department to accurately monitor the general health conditions of the bridges.
The structures have been built to withstand up to a one-minute mean wind speed of 95 metres per second. In 1997, when Hong Kong had a direct hit from Typhoon Victor, wind speeds of 110 to 120 kilometres per hour were recorded. However, the highest wind speed on record occurred during Typhoon Wanda in 1962 when a 3 second gust wind speed was recorded at 78.8 metres per second, 284 kilometres per hour.
The information from these hundreds of different sensors is transmitted to the data acquisition outstation units. There are three data acquisition outstation units on Tsing Ma bridge, three on Ting Kau and two on the Kap Shui Mun.
The computing powerhouse for these systems is in the administrative building used by the Highways Department in Tsing Yi. The local central computer system provides data collection control, post-processing, transmission and storage. The global system is used for data acquisition and analysis, assessing the physical conditions and structural functions of the bridges and for integration and manipulation of the data acquisition, analysis and assessing processes.
The following project are currently known as some of the biggest on-going bridge monitoring
Health monitoring of large bridges shall be performed by simultaneous measurement of loads on the bridge and effects of these loads. It typically includes monitoring of:
Provided with this knowledge, the engineer can:
References are available that provide an introduction to the application of fiber optic sensors to Structural Health Monitoring on bridges.[6]